Observable Changes of Hypotheses under Positive Reinforcement.
نویسندگان
چکیده
In mathematical models of concept learning it has consistently been assumed that positive reinforcement cannot lead to a change of the hypothesis determining the overt response. When hypotheses are experimentally identified and recorded along with positive and negative reinforcements of stimulus-response pairs, it can be shown that hypotheses may change after a positive reinforcement. Positive reinforcement has an information content for subjects that has not yet been adequately recognized in concept formation studies.
منابع مشابه
Reversibility of reinforcement between eating and running by schedule changes: A comparison of hypotheses and models
Rats increased eating that produced access to a running-wheel or increased running that produced access to food. depending on which response was potentially deprived. relative to baseline. by the scheduled ratio of responding. Under both schedules. instrumental responding significantly exceeded appropriate baselines of the noncontingent effects of the schedule. The results contradicted the hypo...
متن کاملMonte Carlo Bayesian Reinforcement Learning
Bayesian reinforcement learning (BRL) encodes prior knowledge of the world in a model and represents uncertainty in model parameters by maintaining a probability distribution over them. This paper presents Monte Carlo BRL (MC-BRL), a simple and general approach to BRL. MC-BRL samples a priori a finite set of hypotheses for the model parameter values and forms a discrete partially observable Mar...
متن کاملReinforcement Learning for Problems with Hidden State
In this paper, we describe how techniques from reinforcement learning might be used to approach the problem of acting under uncertainty. We start by introducing the theory of partially observable Markov decision processes (POMDPs) to describe what we call hidden state problems. After a brief review of other POMDP solution techniques, we motivate reinforcement learning by considering an agent wi...
متن کاملReinforcement Learning for Decentralized Planning Under Uncertainty (Doctoral Consortium)
Decentralized partially-observable Markov decision processes (Dec-POMDPs) are a powerful tool for modeling multi-agent planning and decision-making under uncertainty. Prevalent Dec-POMDP solution techniques require centralized computation given full knowledge of the underlying model. But in real world scenarios, model parameters may not be known a priori, or may be difficult to specify. We prop...
متن کاملModel-Based Probabilistic Pursuit via Inverse Reinforcement Learning
In this paper we address the integrated prediction, planning, and control problem that enables a single follower robot (the “photographer”) to quickly re-establish visual contact with a moving target (the “subject”) that has escaped the follower’s field of view. Our work addresses this unavoidable scenario, which reactive controllers are typically ill-equipped to handle, by making intelligent p...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Science
دوره 148 3670 شماره
صفحات -
تاریخ انتشار 1965